static int vmx_do_page_fault(unsigned long va, struct cpu_user_regs *regs)
{
- struct exec_domain *ed = current;
unsigned long eip;
l1_pgentry_t gpte;
unsigned long gpa; /* FIXME: PAE */
}
#endif
- /*
- * If vpagetable is zero, then we are still emulating 1:1 page tables,
- * and we should have never gotten here.
- */
- if ( !test_bit(VMX_CPU_STATE_PG_ENABLED, &ed->arch.arch_vmx.cpu_state) )
- {
- printk("vmx_do_page_fault while running on 1:1 page table\n");
- return 0;
- }
+ if (!vmx_paging_enabled(current))
+ handle_mmio(va, va);
gpte = gva_to_gpte(va);
if (!(l1e_get_flags(gpte) & _PAGE_PRESENT) )
vio = (vcpu_iodata_t *) d->arch.arch_vmx.vmx_platform.shared_page_va;
if (vio == 0) {
- VMX_DBG_LOG(DBG_LEVEL_1, "bad shared page: %lx", (unsigned long) vio);
+ printk("bad shared page: %lx", (unsigned long) vio);
domain_crash_synchronous();
}
p = &vio->vp_ioreq;
laddr = (p->dir == IOREQ_WRITE) ? regs->esi : regs->edi;
}
p->pdata_valid = 1;
- p->u.pdata = (void *) gva_to_gpa(laddr);
+
+ p->u.data = laddr;
+ if (vmx_paging_enabled(d))
+ p->u.pdata = (void *) gva_to_gpa(p->u.data);
p->df = (eflags & X86_EFLAGS_DF) ? 1 : 0;
if (test_bit(5, &exit_qualification)) /* "rep" prefix */
return 0;
}
- mfn = phys_to_machine_mapping(l1e_get_pfn(gva_to_gpte(laddr)));
+ mfn = phys_to_machine_mapping(laddr >> PAGE_SHIFT);
addr = map_domain_mem((mfn << PAGE_SHIFT) | (laddr & ~PAGE_MASK));
if (dir == COPY_IN)
error |= __vmwrite(CR0_READ_SHADOW, c->cr0);
+ if (!vmx_paging_enabled(d)) {
+ VMX_DBG_LOG(DBG_LEVEL_VMMU, "switching to vmxassist. use phys table");
+ __vmwrite(GUEST_CR3, pagetable_val(d->domain->arch.phys_table));
+ goto skip_cr3;
+ }
+
if (c->cr3 == d->arch.arch_vmx.cpu_cr3) {
/*
* This is simple TLB flush, implying the guest has
*/
mfn = phys_to_machine_mapping(c->cr3 >> PAGE_SHIFT);
if ((mfn << PAGE_SHIFT) != pagetable_val(d->arch.guest_table)) {
- VMX_DBG_LOG(DBG_LEVEL_VMMU, "Invalid CR3 value=%lx", c->cr3);
+ printk("Invalid CR3 value=%lx", c->cr3);
domain_crash_synchronous();
return 0;
}
*/
VMX_DBG_LOG(DBG_LEVEL_VMMU, "CR3 c->cr3 = %lx", c->cr3);
if ((c->cr3 >> PAGE_SHIFT) > d->domain->max_pages) {
- VMX_DBG_LOG(DBG_LEVEL_VMMU, "Invalid CR3 value=%lx", c->cr3);
+ printk("Invalid CR3 value=%lx", c->cr3);
domain_crash_synchronous();
return 0;
}
__vmwrite(GUEST_CR3, pagetable_val(d->arch.shadow_table));
}
+skip_cr3:
+
error |= __vmread(CR4_READ_SHADOW, &old_cr4);
error |= __vmwrite(GUEST_CR4, (c->cr4 | X86_CR4_VMXE));
error |= __vmwrite(CR4_READ_SHADOW, c->cr4);
struct exec_domain *d = current;
unsigned long old_base_mfn, mfn;
unsigned long eip;
+ int paging_enabled;
/*
* CR0: We don't want to lose PE and PG.
*/
+ paging_enabled = vmx_paging_enabled(d);
__vmwrite(GUEST_CR0, (value | X86_CR0_PE | X86_CR0_PG));
+ __vmwrite(CR0_READ_SHADOW, value);
- if (value & (X86_CR0_PE | X86_CR0_PG) &&
- !test_bit(VMX_CPU_STATE_PG_ENABLED, &d->arch.arch_vmx.cpu_state)) {
- /*
- * Enable paging
- */
- set_bit(VMX_CPU_STATE_PG_ENABLED, &d->arch.arch_vmx.cpu_state);
+ VMX_DBG_LOG(DBG_LEVEL_VMMU, "Update CR0 value = %lx\n", value);
+ if ((value & X86_CR0_PE) && (value & X86_CR0_PG)
+ && !paging_enabled) {
/*
* The guest CR3 must be pointing to the guest physical.
*/
d->arch.arch_vmx.cpu_cr3 >> PAGE_SHIFT)) ||
!get_page(pfn_to_page(mfn), d->domain) )
{
- VMX_DBG_LOG(DBG_LEVEL_VMMU, "Invalid CR3 value = %lx",
- d->arch.arch_vmx.cpu_cr3);
+ printk("Invalid CR3 value = %lx", d->arch.arch_vmx.cpu_cr3);
domain_crash_synchronous(); /* need to take a clean path */
}
old_base_mfn = pagetable_get_pfn(d->arch.guest_table);
} else {
if ((value & X86_CR0_PE) == 0) {
__vmread(GUEST_EIP, &eip);
- VMX_DBG_LOG(DBG_LEVEL_1,
- "Disabling CR0.PE at %%eip 0x%lx", eip);
+ VMX_DBG_LOG(DBG_LEVEL_1, "Disabling CR0.PE at %%eip 0x%lx\n", eip);
if (vmx_assist(d, VMX_ASSIST_INVOKE)) {
set_bit(VMX_CPU_STATE_ASSIST_ENABLED,
&d->arch.arch_vmx.cpu_state);
switch(cr) {
case 0:
{
- __vmwrite(CR0_READ_SHADOW, value);
return vmx_set_cr0(value);
}
case 3:
/*
* If paging is not enabled yet, simply copy the value to CR3.
*/
- if (!test_bit(VMX_CPU_STATE_PG_ENABLED, &d->arch.arch_vmx.cpu_state)) {
+ if (!vmx_paging_enabled(d)) {
d->arch.arch_vmx.cpu_cr3 = value;
break;
}
!VALID_MFN(mfn = phys_to_machine_mapping(value >> PAGE_SHIFT)) ||
!get_page(pfn_to_page(mfn), d->domain) )
{
- VMX_DBG_LOG(DBG_LEVEL_VMMU,
- "Invalid CR3 value=%lx", value);
+ printk("Invalid CR3 value=%lx", value);
domain_crash_synchronous(); /* need to take a clean path */
}
old_base_mfn = pagetable_get_pfn(d->arch.guest_table);
VMX_DBG_LOG(DBG_LEVEL_0, "exit reason = %x", exit_reason);
if (exit_reason & VMX_EXIT_REASONS_FAILED_VMENTRY) {
+ printk("Failed vm entry\n");
domain_crash_synchronous();
return;
}